Interactive Agents to Overcome Ambiguity in Software Engineering
In this work, we study the ability of LLM agents to handle ambiguous instructions in interactive code generation settings by evaluating proprietary and open-weight models on their performance across three key steps: (a) leveraging interactivity to improve performance in ambiguous scenarios, (b) detecting ambiguity, and (c) asking targeted questions.
曖昧な指示にエージェントが質問するベンチマーク